Exploring Non-target Knowledge for Improving Ensemble Universal Adversarial Attacks

نویسندگان

چکیده

The ensemble attack with average weights can be leveraged for increasing the transferability of universal adversarial perturbation (UAP) by training multiple Convolutional Neural Networks (CNNs). However, after analyzing Pearson Correlation Coefficients (PCCs) between logits and individual crafted UAP trained attack, we find that one CNN plays a dominant role during optimization. Consequently, this weighted strategy will weaken contributions other CNNs thus limit black-box CNNs. To deal bias issue, primary attempt is to leverage Kullback–Leibler (KL) divergence loss encourage joint contribution from different CNNs, which still insufficient. After decoupling KL into target-class part non-target-class part, main issue lies in non-target knowledge significantly suppressed due logit target class. In study, simply adopt only considers classes addressing issue. Besides, further boost transferability, incorporate min-max learning framework self-adjust each CNN. Experiments results validate considering achieve superior than original large margin, provide mutual benefit attacks. source code available at: https://github.com/WJJLL/ND-MM.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Ensemble Adversarial Training: Attacks and Defenses

Machine learning models are vulnerable to adversarial examples, inputs maliciously perturbed to mislead the model. These inputs transfer between models, thus enabling black-box attacks against deployed models. Adversarial training increases robustness to attacks by injecting adversarial examples into training data. Surprisingly, we find that although adversarially trained models exhibit strong ...

متن کامل

Defending Non-Bayesian Learning against Adversarial Attacks

Abstract This paper addresses the problem of non-Bayesian learning over multi-agent networks, where agents repeatedly collect partially informative observations about an unknown state of the world, and try to collaboratively learn the true state. We focus on the impact of the adversarial agents on the performance of consensus-based non-Bayesian learning, where non-faulty agents combine local le...

متن کامل

Improving Network Robustness against Adversarial Attacks with Compact Convolution

Though Convolutional Neural Networks (CNNs) have surpassed human-level performance on tasks such as object classification and face verification, they can easily be fooled by adversarial attacks. These attacks add a small perturbation to the input image that causes the network to mis-classify the sample. In this paper, we focus on neutralizing adversarial attacks by compact feature learning. In ...

متن کامل

Improving DNN Robustness to Adversarial Attacks using Jacobian Regularization

Deep neural networks have lately shown tremendous performance in various applications including vision and speech processing tasks. However, alongside their ability to perform these tasks with such high accuracy, it has been shown that they are highly susceptible to adversarial attacks: a small change of the input would cause the network to err with high confidence. This phenomenon exposes an i...

متن کامل

Adversarial Attacks on Image Recognition

The purpose of this project is to extend the work done by Papernot et al. in [4] on adversarial attacks in image recognition. We investigated whether a reduction in feature dimensionality can maintain a comparable level of misclassification success while increasing computational efficiency. We formed an attack on a black-box model with an unknown training set by forcing the oracle to misclassif...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2023

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v37i3.25377